Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FIX: Set fs.automatic.close to false in Hadoop configuration #614

Merged
merged 1 commit into from
Apr 28, 2022

Conversation

snehashisp
Copy link
Member

Problem

The Hadoop DFS client auto closes and created clients, when the JVM shuts down. This can happen prior to the connector being closed, which requires these clients to be in open state for deletion of temporary files on close. Note that this error only comes into the picture when the connect worker is stopped/restarted.

Solution

Hadoop has a configuration fs.automatic.close which is by default true. Setting this as false will disable this behaviour. This is safe to do as we are anyways not relying on the auto close feature for closing all the clients. The connector lifecycle has appropriate routines for closing these clients on exit since it needs to rely on the lifecycle during a a connector delete, where the JVM shut down hooks are not executed. The connector should always operate with fs.automatic.close as false.

Does this solution apply anywhere else?
  • yes
  • no
If yes, where?

Test Strategy

Testing done:
  • Unit tests
  • Integration tests
  • System tests
  • Manual tests

Release Plan

@snehashisp snehashisp requested a review from a team as a code owner April 27, 2022 14:59
Copy link
Member

@kpatelatwork kpatelatwork left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants